Convolution Aware Initialization

نویسنده

  • Armen Aghajanyan
چکیده

Initialization of parameters in deep neural networks has been shown to have a big impact on the performance of the networks (Mishkin & Matas, 2015). The initialization scheme devised by He et al, allowed convolution activations to carry a constrained mean which allowed deep networks to be trained effectively (He et al., 2015a). Orthogonal initializations and more generally orthogonal matrices in standard recurrent networks have been proved to eradicate the vanishing and exploding gradient problem (Pascanu et al., 2012). Majority of current initialization schemes do not take fully into account the intrinsic structure of the convolution operator. This paper introduces a new type of initialization built around the duality of the Fourier transform and the convolution operator. With Convolution Aware Initialization we noticed not only higher accuracy and lower loss, but faster convergence in general. We achieve new state of the art on the CIFAR10 dataset, and achieve close to state of the art on various other tasks.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Fast initialization of Nyquist echo cancelers using circular convolution technique

For full-duplex high-speed data transmission over the two-wire line using the same frequency band, it is required to sufficiently suppress the echo. The use of a conventional adaptation method may take a long time to train the echo canceler. Fast training can be achieved by initializing the coefficients of the echo canceler with an estimate of the impulse response of the echo path. In this lett...

متن کامل

A Non-Convex Optimization Technique for Sparse Blind Deconvolution -- Initialization Aspects and Error Reduction Properties

Sparse blind deconvolution is the problem of estimating the blur kernel and sparse excitation, both of which are unknown. Considering a linear convolution model, as opposed to the standard circular convolution model, we derive a sufficient condition for stable deconvolution. The columns of the linear convolution matrix form a Riesz basis with the tightness of the Riesz bounds determined by the ...

متن کامل

All you need is a good init

Layer-sequential unit-variance (LSUV) initialization – a simple method for weightinitialization for deep net learning – is proposed. The method consists of the twosteps. First, pre-initialize weights of each convolution or inner-product layer withorthonormal matrices. Second, proceed from the first to the final layer, normaliz-ing the variance of the output of each layer to be e...

متن کامل

Automatic Medical Image Segmentation Based on Vfc-snake

An automatic approach to contour segmentation of Computed Tomography (CT) images is presented in this work. Image segmentation is achieved by means of the snake algorithm and the dynamic programming (DP) optimization technique. Based upon the Vector field convolution (VFC), a new strategy for contour points initialization and splitting is presented. Contour initialization is carried out from VF...

متن کامل

Compression Artifacts Removal Using Convolutional Neural Networks

This paper shows that it is possible to train large and deep convolutional neural networks (CNN) for JPEG compression artifacts reduction, and that such networks can provide significantly better reconstruction quality compared to previously used smaller networks as well as to any other state-of-the-art methods. We were able to train networks with 8 layers in a single step and in relatively shor...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1702.06295  شماره 

صفحات  -

تاریخ انتشار 2017